# DPO Fine-Tuning Optimization
Gemma3 27b Abliterated Dpo
A fine-tuned large language model based on mlabonne/gemma-3-27b-it-abliterated, trained using the Unsloth acceleration framework and Huggingface's TRL library, achieving a 2x efficiency improvement.
Large Language Model
Transformers English

G
summykai
326
3
Chocolatine Fusion 14B
MIT
Chocolatine-Fusion-14B is a merged model that combines the strengths of the Chocolatine-2 series, enhancing reasoning capabilities and multi-turn dialogue performance through optimized fusion.
Large Language Model
Transformers

C
FINGU-AI
226
5
Neuralbeagle14 7B 8.0bpw H8 Exl2
Apache-2.0
NeuralBeagle14-7B is a 7B-parameter large language model fine-tuned using the DPO method based on the Beagle14-7B model, excelling in the 7B parameter category.
Large Language Model
Transformers

N
LoneStriker
111
5
Featured Recommended AI Models